Attention Flows: Analyzing and Comparing Attention Mechanisms in Language Models

نویسندگان

چکیده

Advances in language modeling have led to the development of deep attention-based models that are performant across a wide variety natural processing (NLP) problems. These typified by pre-training process on large unlabeled text corpora and subsequently fine-tuned for specific tasks. Although considerable work has been devoted understanding attention mechanisms pre-trained models, it is less understood how model's change when trained target NLP task. In this paper, we propose visual analytics approach fine-tuning models. Our visualization, Attention Flows, designed support users querying, tracing, comparing within layers, amongst heads Transformer-based To help gain insight classification decision made, our design centered depicting classification-based at deepest layer from prior layers flows throughout words input. Flows supports analysis single model, as well comparison between via their similarities differences. We use study various sentence tasks highlight evolves address nuances solving these

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Comparing the Relationship between Focus of Attention and Attention Bias in Suicide and Non-suicide Attempters

Objective: Suicide rate is unfortunately increasing in recent years and its assessment is a must. Most studies have evaluated different factors like personality, depression, or environment in suicide attempters but cognitive functions have been overlooked in suicide attempters. The current research aimed to evaluate a kind of cognitive function, namely attention in survivors of suicide.  ...

متن کامل

Attention and Visual Perception: Analyzing Divided Attention

Attention is one of the most active areas of experimental research in cognitive science. This chapter focuses on an aspect of attention that has been studied intensively over the past twenty-five years or so: capacity limitations on our ability to perceive multiple visual inputs. The basic questions are simple but far-reaching. How much visual information can we take in at one time? What can we...

متن کامل

Coherent Dialogue with Attention-Based Language Models

We model coherent conversation continuation via RNNbased dialogue models equipped with a dynamic attention mechanism. Our attention-RNN language model dynamically increases the scope of attention on the history as the conversation continues, as opposed to standard attention (or alignment) models with a fixed input scope in a sequence-tosequence model. This allows each generated word to be assoc...

متن کامل

Mechanisms and Representations of Language-Mediated Visual Attention

The experimental investigation of language-mediated visual attention is a promising way to study the interaction of the cognitive systems involved in language, vision, attention, and memory. Here we highlight four challenges for a mechanistic account of this oculomotor behavior: the levels of representation at which language-derived and vision-derived representations are integrated; attentional...

متن کامل

comparing the relationship between focus of attention and attention bias in suicide and non-suicide attempters

objective: suicide rate is unfortunately increasing in recent years and its assessment is a must. most studies have evaluated different factors like personality, depression, or environment in suicide attempters but cognitive functions have been overlooked in suicide attempters. the current research aimed to evaluate a kind of cognitive function, namely attention in survivors of suicide.  method...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Visualization and Computer Graphics

سال: 2021

ISSN: ['1077-2626', '2160-9306', '1941-0506']

DOI: https://doi.org/10.1109/tvcg.2020.3028976